首页> 外文OA文献 >Generating Adversarial Examples with Adversarial Networks
【2h】

Generating Adversarial Examples with Adversarial Networks

机译:用对抗网络生成对抗性实例

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Deep neural networks (DNNs) have been found to be vulnerable to adversarialexamples resulting from adding small-magnitude perturbations to inputs. Suchadversarial examples can mislead DNNs to produce adversary-selected results.Different attack strategies have been proposed to generate adversarialexamples, but how to produce them with high perceptual quality and moreefficiently requires more research efforts. In this paper, we propose AdvGAN togenerate adversarial examples with generative adversarial networks (GANs),which can learn and approximate the distribution of original instances. ForAdvGAN, once the generator is trained, it can generate adversarialperturbations efficiently for any instance, so as to potentially accelerateadversarial training as defenses. We apply AdvGAN in both semi-whitebox andblack-box attack settings. In semi-whitebox attacks, there is no need to accessthe original target model after the generator is trained, in contrast totraditional white-box attacks. In black-box attacks, we dynamically train adistilled model for the black-box model and optimize the generator accordingly.Adversarial examples generated by AdvGAN on different target models have highattack success rate under state-of-the-art defenses compared to other attacks.Our attack has placed the first with 92.76% accuracy on a public MNISTblack-box attack challenge.
机译:深度神经网络(DNN)被发现容易受到对抗示例的攻击,这些示例是由于向输入添加小幅度扰动而产生的。这样的对抗性例子会误导DNN产生对手选择的结果。已提出了不同的攻击策略来生成对抗性例子,但是如何以高感知质量和更有效地产生它们却需要更多的研究工作。在本文中,我们建议使用AdvGAN来生成具有生成对抗网络(GAN)的对抗示例,该网络可以学习并近似原始实例的分布。对于AdGAN,一旦训练了生成器,就可以在任何情况下有效地生成对抗性扰动,从而有可能加速对抗性防御的防御。我们在半白盒和黑盒攻击设置中都应用了AdvGAN。在半白盒攻击中,与传统的白盒攻击相比,在训练生成器后无需访问原始目标模型。在黑盒攻击中,我们为黑盒模型动态训练了蒸馏模型并相应地优化了生成器。与其他攻击相比,AdvGAN在不同目标模型上生成的示例实例具有较高的攻击成功率。我们的攻击在公开的MNISTblack-box攻击挑战中以92.76%的准确度排名第一。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号